138 research outputs found
A Corpus for Hybrid Question Answering Systems
International audienceQuestion answering has been the focus of a lot of researches and evaluation campaigns, either for text-based systems (TREC and CLEF evaluation campaigns for example), or for knowledge-based systems (QALD, BioASQ). Few systems have effectively combined both types of resources and methods in order to exploit the fruitful- ness of merging the two kinds of information repositories. The only evaluation QA track that focuses on hybrid QA is QALD since 2014. As it is a recent task, few annotated data are available (around 150 questions). In this paper, we present a question answering dataset that was constructed to develop and evaluate hybrid question an- swering systems. In order to create this corpus, we collected several textual corpora and augmented them with entities and relations of a knowledge base by retrieving paths in the knowledge base which allow to answer the questions. The resulting corpus contains 4300 question-answer pairs and 1600 have a true link with DBpedia
Morphological resources for precise information retrieval
International audienceQuestion answering (QA) systems aim at providing a precise answer to a given user question. Their major difficulty lies in the lexical gap problem between question and answering passages. We present here the different types of morphological phenomena in question answering, the resources available for French, and in particular a resource that we built containing deverbal agent nouns. Then, we evaluate the results of a particular QA system, according to the morphological knowledge used
Corpus study of kidney-related experimental data in scientific papers
International audienceThe Quantitative Kidney DataBase (QKDB) is relational database that was created in order to centralize kidney-related experimental results. Each result is characterized by different attributes and the scientific paper from which it was extracted. Currently, this database isp opulated by hand by experts of the domain. We present a corpus study of some papers that have already been analyzed in order to exhibit the specificities and difficulties of the extraction process; then we propose a first solution to extract automatically the desired knowledge from papers
Fine-grained Linguistic Evaluation of Question Answering Systems
International audienceQuestion answering systems are complex systems using natural language processing. Some evaluation campaigns are organized to evaluate such systems in order to propose a classification of systems based on final results (number of correct answers). Nevertheless, teams need to evaluate more precisely the results obtained by their systems if they want to do a diagnostic evaluation. There are no tools or methods to do these evaluations systematically. We present REVISE, a tool for glass box evaluation based on diagnostic of question answering system results
Feature selection for drug-drug interaction detection using machine-learning based approaches
International audienceThis paper describes the systems developed for the DDI Extraction challenge. The systems use machine learning methods and are based on SVM by using LIBSVM and SVMPerf tools. Classical features and corpus-specific features are selected according to their F-score. The best system obtained an F-measure of 0.5965
Es esch fĂ scht wie Ditsch, oder net? Ătiquetage morphosyntaxique de l'alsacien en passant par l'allemand
National audienceLâanalyse morphosyntaxique est une pierre angulaire de nombreuses applications du traitement automatique des langues. Elle nĂ©cessite toutefois des efforts de dĂ©veloppement consĂ©quents, quâil sâagisse dâannoter des corpus ou de produire des lexiques et des outils. Pour les langues moins dotĂ©es, il peut ĂȘtre utile dâexploiter leur proximitĂ© avec des langues disposant dâoutils et de ressources. Dans cet article, nous nous intĂ©ressons plus particuliĂšrement aux dialectes alsaciens, qui prĂ©sentent de nombreuses similitudes avec lâallemand standard. Nous montrons quâil est possible dâutiliser des outils dĂ©veloppĂ©s pour lâallemand afin de rĂ©aliser lâanalyse morphosyntaxique de textes en alsacien. La mĂ©thode consiste Ă transposer les mots outils des textes alsaciens vers leurs Ă©quivalents en allemand standard. Cette transposition nĂ©cessite pour seule ressource un lexiquebilingue des mots outils
FRASQUES : A Question-Answering System in the EQueR Evaluation Campaign
Ă paraĂźtreInternational audienceQuestion-answering (QA) systems aim at providing either a small passage or just the answer to a question in natural language. We have developed several QA systems that work on both English and French. This way, we are able to provide answers to questions given in both languages by searching documents in both languages also. In this article, we present our French monolingual system FRASQUES which participated in the EQueR evaluation campaign of QA systems for French in 2004. First, the QA architecture common to our systems is shown. Then, for every step of the QA process, we consider which steps are language-independent, and for those that are language-dependent, the tools or processes that need to be adapted to switch for one language to another. Finally, our results at EQueR are given and commented; an error analysis is conducted, and the kind of knowledge needed to answer a question is studied
Semantic knowledge in Question-Answering systems
International audienceQA systems need semantic knowledge to find in documents variations of the question terms. They benefit from the use of knowledge resources such as synonym dictionaries or ontologies like WordNet. Our goal here is to study to which extent variations are needed and to determine what kinds of variations are useful or necessary for these systems. This study is based on different corpora in which we analyze semantic term variations, based on reference sets of possible variations
Lexical validation of answers in question answering
International audienceQuestion answering (QA) aims at retrieving precise information from a large collection of documents, typically the Web. Different techniques can be used to find relevant information, and to compare these techniques, it is important to evaluate question answering systems. The objective of an Answer Validation task is to estimate the correctness of an answer returned by a QA system for a question, according to the text snippet given to support it. We participated in such a task in 2006. In this article, we present our strategy for deciding if the snippets justify the answers. We used a strategy based on our own question answering system, and compared the answers it returned with the answer to judge. We discuss our results, and show the possible extensions of our strategy. Then we point out the difficulties of this task, by examining different examples
- âŠ